Distance metric learning (DML) is an important task that has foundapplications in many domains. The high computational cost of DML arises fromthe large number of variables to be determined and the constraint that adistance metric has to be a positive semi-definite (PSD) matrix. Althoughstochastic gradient descent (SGD) has been successfully applied to improve theefficiency of DML, it can still be computationally expensive because in orderto ensure that the solution is a PSD matrix, it has to, at every iteration,project the updated distance metric onto the PSD cone, an expensive operation.We address this challenge by developing two strategies within SGD, i.e.mini-batch and adaptive sampling, to effectively reduce the number of updates(i.e., projections onto the PSD cone) in SGD. We also develop hybrid approachesthat combine the strength of adaptive sampling with that of mini-batch onlinelearning techniques to further improve the computational efficiency of SGD forDML. We prove the theoretical guarantees for both adaptive sampling andmini-batch based approaches for DML. We also conduct an extensive empiricalstudy to verify the effectiveness of the proposed algorithms for DML.
展开▼
机译:距离度量学习(DML)是一项重要任务,已在许多领域中找到了应用。 DML的高计算成本来自要确定的大量变量以及距离度量必须为正半定(PSD)矩阵的约束。尽管随机梯度下降(SGD)已成功用于提高DML的效率,但是它仍然在计算上昂贵,因为为了确保解决方案是PSD矩阵,它必须在每次迭代时将更新的距离度量投影到PSD上我们通过在SGD中开发两种策略(即小批量和自适应采样)来有效减少SGD中的更新次数(即投影到PSD锥上的投影),从而解决了这一挑战。我们还开发了混合方法,将自适应采样的强度与小批量在线学习技术的强度相结合,以进一步提高SDM for DML的计算效率。我们证明了DML的自适应采样和基于小批量方法的理论保证。我们还进行了广泛的经验研究,以验证所提出的DML算法的有效性。
展开▼